抗微生物抗性(AMR)是患者的风险和医疗保健系统的负担。但是,AMR测定通常需要几天。本研究为基于易于使用的临床和微生物预测因子,包括患者人口统计,医院住宿数据,诊断,临床特征以及微生物/抗微生物特征,以及仅使用微生物/抗微生物特征将这些模型与微生物/抗微生物特性进行基于幼稚抗体模型的模型的预测模型。在培养之前准确地预测阻力的能力可以向临床决策提供通知临床决策并缩短行动时间。这里采用的机器学习算法显示出改进的分类性能(接收器操作特性曲线0.88-0.89的区域)与使用飞利浦EICU研究所的6个生物和10个抗生素的接收器操作特征曲线0.86下的接收器下的面积为0.88-0.89)(ERI )数据库。该方法可以帮助指导抗菌治疗,目的是改善患者结果并减少不必要或无效抗生素的使用。
translated by 谷歌翻译
Temporal reasoning is the task of predicting temporal relations of event pairs with corresponding contexts. While some temporal reasoning models perform reasonably well on in-domain benchmarks, we have little idea of the systems' generalizability due to existing datasets' limitations. In this work, we introduce a novel task named TODAY that bridges this gap with temporal differential analysis, which as the name suggests, evaluates if systems can correctly understand the effect of incremental changes. Specifically, TODAY makes slight context changes for given event pairs, and systems need to tell how this subtle contextual change will affect temporal relation distributions. To facilitate learning, TODAY also annotates human explanations. We show that existing models, including GPT-3, drop to random guessing on TODAY, suggesting that they heavily rely on spurious information rather than proper reasoning for temporal predictions. On the other hand, we show that TODAY's supervision style and explanation annotations can be used in joint learning and encourage models to use more appropriate signals during training and outperform across several benchmarks. TODAY can also be used to train models to solicit incidental supervision from noisy sources such as GPT-3 and moves farther towards generic temporal reasoning systems.
translated by 谷歌翻译
As the size of the dataset used in deep learning tasks increases, the noisy label problem, which is a task of making deep learning robust to the incorrectly labeled data, has become an important task. In this paper, we propose a method of learning noisy label data using the label noise selection with test-time augmentation (TTA) cross-entropy and classifier learning with the NoiseMix method. In the label noise selection, we propose TTA cross-entropy by measuring the cross-entropy to predict the test-time augmented training data. In the classifier learning, we propose the NoiseMix method based on MixUp and BalancedMix methods by mixing the samples from the noisy and the clean label data. In experiments on the ISIC-18 public skin lesion diagnosis dataset, the proposed TTA cross-entropy outperformed the conventional cross-entropy and the TTA uncertainty in detecting label noise data in the label noise selection process. Moreover, the proposed NoiseMix not only outperformed the state-of-the-art methods in the classification performance but also showed the most robustness to the label noise in the classifier learning.
translated by 谷歌翻译
Uncertainty estimation of the trained deep learning network provides important information for improving the learning efficiency or evaluating the reliability of the network prediction. In this paper, we propose a method for the uncertainty estimation for multi-class image classification using test-time mixup augmentation (TTMA). To improve the discrimination ability between the correct and incorrect prediction of the existing aleatoric uncertainty, we propose the data uncertainty by applying the mixup augmentation on the test data and measuring the entropy of the histogram of predicted labels. In addition to the data uncertainty, we propose a class-specific uncertainty presenting the aleatoric uncertainty associated with the specific class, which can provide information on the class confusion and class similarity of the trained network. The proposed methods are validated on two public datasets, the ISIC-18 skin lesion diagnosis dataset, and the CIFAR-100 real-world image classification dataset. The experiments demonstrate that (1) the proposed data uncertainty better separates the correct and incorrect prediction than the existing uncertainty measures thanks to the mixup perturbation, and (2) the proposed class-specific uncertainty provides information on the class confusion and class similarity of the trained network for both datasets.
translated by 谷歌翻译
This paper examines the encoding of analogy in large-scale pretrained language models, such as BERT and GPT-2. Existing analogy datasets typically focus on a limited set of analogical relations, with a high similarity of the two domains between which the analogy holds. As a more realistic setup, we introduce the Scientific and Creative Analogy dataset (SCAN), a novel analogy dataset containing systematic mappings of multiple attributes and relational structures across dissimilar domains. Using this dataset, we test the analogical reasoning capabilities of several widely-used pretrained language models (LMs). We find that state-of-the-art LMs achieve low performance on these complex analogy tasks, highlighting the challenges still posed by analogy understanding.
translated by 谷歌翻译
Rates of missing data often depend on record-keeping policies and thus may change across times and locations, even when the underlying features are comparatively stable. In this paper, we introduce the problem of Domain Adaptation under Missingness Shift (DAMS). Here, (labeled) source data and (unlabeled) target data would be exchangeable but for different missing data mechanisms. We show that when missing data indicators are available, DAMS can reduce to covariate shift. Focusing on the setting where missing data indicators are absent, we establish the following theoretical results for underreporting completely at random: (i) covariate shift is violated (adaptation is required); (ii) the optimal source predictor can perform worse on the target domain than a constant one; (iii) the optimal target predictor can be identified, even when the missingness rates themselves are not; and (iv) for linear models, a simple analytic adjustment yields consistent estimates of the optimal target parameters. In experiments on synthetic and semi-synthetic data, we demonstrate the promise of our methods when assumptions hold. Finally, we discuss a rich family of future extensions.
translated by 谷歌翻译
It has been experimentally demonstrated that humans are able to learn in a manner that allows them to make predictions on categories for which they have not seen any examples (Malaviya et al., 2022). Sucholutsky and Schonlau (2020) have recently presented a machine learning approach that aims to do the same. They utilise synthetically generated data and demonstrate that it is possible to achieve sub-linear scaling and develop models that can learn to recognise N classes from M training samples where M is less than N - aka less-than-one shot learning. Their method was, however, defined for univariate or simple multivariate data (Sucholutsky et al., 2021). We extend it to work on large, high-dimensional and real-world datasets and empirically validate it in this new and challenging setting. We apply this method to learn previously unseen NLP tasks from very few examples (4, 8 or 16). We first generate compact, sophisticated less-than-one shot representations called soft-label prototypes which are fitted on training data, capturing the distribution of different classes across the input domain space. We then use a modified k-Nearest Neighbours classifier to demonstrate that soft-label prototypes can classify data competitively, even outperforming much more computationally complex few-shot learning methods.
translated by 谷歌翻译
新的AUV技术的发展增加了AUV可以应对的任务范围及其运营的长度。结果,AUV能够处理高度复杂的操作。但是,这些任务并不容易适合将任务定义为一系列预先计划的航路点的传统方法,因为不可能事先知道,在任务过程中可能发生的一切。这会导致操作员的期望和实际操作绩效之间存在差距。因此,这可能会在操作员和AUV之间产生降低的信任程度,从而导致不必要的任务中断。为了弥合机器人行为和运营商的期望之间的这一差距,这项工作旨在提供一个框架,以易于理解的方式解释任务期间自动驾驶汽车采取的决策和行动。此外,目的是拥有一个自治性系统,可以在任何自治体系结构之上添加为附加层。为了使该方法适用于配备不同自主权的不同自主系统,这项工作将自主权的内部运作与决策点以及应用知识蒸馏的由此产生的执行动作。最后,为了以更自然的方式向操作员介绍解释,蒸馏决策树的输出与自然语言解释相结合,并将其报告给操作员作为句子。因此,在解释管道的末尾添加了一个称为Concept2Text生成的附加步骤。
translated by 谷歌翻译
最先进的(SOTA)深度学习乳房X线照片分类器接受了弱标记的图像训练,通常依赖于产生有限解释性预测的全球模型,这是他们成功地转化为临床实践的关键障碍。另一方面,基于原型的模型通过将预测与训练图像原型相关联,改善了可解释性,但是它们的准确性不如全球模型,其原型往往具有差的多样性。我们通过BraixProtopnet ++的建议解决了这两个问题,该问题通过将基于原型的模型结合起来,为全局模型增添了解释性。 BraixProtopnet ++在训练基于原型的模型以提高合奏的分类精度时,会提炼全局模型的知识。此外,我们提出了一种方法来通过保证所有原型都与不同的训练图像相关联,以增加原型多样性。对弱标记的私人和公共数据集进行的实验表明,BraixProtopnet ++的分类精度比基于SOTA Global和基于原型的模型具有更高的分类精度。使用病变定位来评估模型可解释性,我们显示BraixProtopnet ++比其他基于原型的模型和全球模型的事后解释更有效。最后,我们表明,BraixProtopnet ++学到的原型的多样性优于基于SOTA原型的方法。
translated by 谷歌翻译
我们提出了一种多阶段的多代码书(MSMC)方法,用于高性能神经TTS合成。基于矢量定量的,变异的自动编码器(VQ-VAE)的特征分析仪用于编码语音训练数据的MEL频谱图,通过在多个阶段中逐渐减小为MSMC表示(MSMCR),并使用不同的时间分辨率进行逐步降低,并使用多个VQ对其进行量化它们代码书分别。通过最大程度地减少重建均方根误差(MSE)和“三重态损耗”的合并损失,对多阶段预测指标进行了训练,以逐步将输入文本序列映射到MSMCR。在合成中,神经声码器将预测的MSMCR转换为最终的语音波形。拟议的方法是由女演讲者通过16小时的英语TTS数据库进行了训练和测试。拟议的TTS的MOS得分为4.41,其表现优于基线,MOS为3.62。拟议的TTS的紧凑版本仍然可以保留高MOS得分。消融研究表明,多个阶段和多个代码手册都可以有效地实现高TTS性能。
translated by 谷歌翻译